Uncertainty beyond sampling error.
نویسندگان
چکیده
Statistical analysis of research results mainly uses confidence intervals and hypothesis tests to capture the uncertainty rising from our study being on a sample of participants drawn from a much larger population, in which our interest mainly lies. But beyond the issue of sampling variation there are other sources of uncertainty that may be even more important to consider. In measurement, a distinction is made between precision, which is how variable are measurements on the same person by the same method made at the same time, and accuracy, which is how close the measurement is to what we actually want to know. For example, if we were to ask a group of patients on two occasions howmuch alcohol they typically consume, this would enable us to estimate precision, how repeatable answers are, but not how close these answers are to how much they actually drink, which we might suspect to be higher. In the same way, a confidence interval tells us about the precision of research results, what would happen if we were to repeat the same study, not their accuracy, which is how close the study is to the truth. In general, beyond the imprecision or uncertainty of numerical results arising from sampling, the main concern is the possibility that the study results are biased. Recent developments in appraising published randomised trials have switched from considering “quality” (essentially undefinable) to assessing explicitly the risk of bias in relation to the way the study was done. Here sources of possible bias include lack of blinding and losses to follow up (missing data). But bias is especially relevant in non-randomised studies, where there will be major concerns about possible confounding, where an apparent relationship between two things may be the result of the relationships of both to a third. A further source of uncertainty concerns the extent to which the results of research conducted in a particular setting with selected participants can be taken as applying equally to a wider group of patients in a different location. Judgement of generalisability (also known as external validity) is challenging. In a clinical trial, the part of the larger population we particularly care about does not yet exist. We want to know what would happen to future patients if we were to apply either of the trial treatments. Changes over time in the nature of disease, in the fitness of the patient, in nutrition, in ancillary treatments, and so on, may all make our sample unrepresentative as a guide for future action. For example, the UK review of evidence relating to mammographic screening outlined three sources of uncertainty in relation to the pooled estimated effect from a meta-analysis of the results of all the randomised trials. First, there was uncertainty due to sampling variation, as previously discussed. Second, there was uncertainty from some methodological weaknesses of the trials. Third was uncertainty about whether the results from the trials were still relevant 25 years later, after major changes in cancer incidence, management of and treatments for breast cancer, and the technology of mammography. Unlike sampling variation, which is quantified in the confidence interval and is uncontroversial, the uncertainty from the other causes cannot easily be quantified and remains the source of fierce debate. Large numbers, increasingly common in this era of big data, will produce narrow confidence intervals. These can create an illusion of accuracy, but they ignore all sources of possible bias that are not affected by sample size, and so these other sources become relatively more important. A recent example of a very precise but seriously wrong answer purported to show that skin cancer was protective for heart attacks and all-cause mortality. So, although confidence intervals are a valuable way of depicting uncertainty, they are always too narrow in the sense that they reflect only statistical uncertainty, precision rather than accuracy. Many journals require authors to consider in their discussion the limitations of their study—some even require this in the article’s abstract. Issues raised there help readers to judge what extra uncertainty might apply to the study, including whether observed effects may be affected by bias. It is common for authors to say that their results should be “interpreted with caution” (including >700 in the BMJ), but who knows what that means in practice? The GRADE group have developed a framework for a more structured approach to assessing the reliability of research findings that addresses the aspects outlined above.
منابع مشابه
Error and Uncertainty Quantification and Sensitivity Analysis Inmechanics Computationalmodels
Multiple sources of errors and uncertainty arise in mechanics computational models and contribute to the uncertainty in the final model prediction. This paper develops a systematic error quantification methodology for computational models. Some types of errors are deterministic, and some are stochastic. Appropriate procedures are developed to either correct the model prediction for deterministi...
متن کاملQuantifying errors without random sampling
BACKGROUND All quantifications of mortality, morbidity, and other health measures involve numerous sources of error. The routine quantification of random sampling error makes it easy to forget that other sources of error can and should be quantified. When a quantification does not involve sampling, error is almost never quantified and results are often reported in ways that dramatically oversta...
متن کاملFoundations of Probability
Probability theory is that part of mathematics that is concerned with the description and modeling of random phenomena, or in a more general — but not unanimously accepted — sense, of any kind of uncertainty. Probability is assigned to random events, expressing their tendency to occur in a random experiment, or more generally to propositions, characterizing the degree of belief in their truth. ...
متن کاملA Bootstrap Interval Robust Data Envelopment Analysis for Estimate Efficiency and Ranking Hospitals
Data envelopment analysis (DEA) is one of non-parametric methods for evaluating efficiency of each unit. Limited resources in healthcare economy is the main reason in measuring efficiency of hospitals. In this study, a bootstrap interval data envelopment analysis (BIRDEA) is proposed for measuring the efficiency of hospitals affiliated with the Hamedan University of Medical Sciences. The propos...
متن کاملHeterogenic Solid Biofuel Sampling Methodology and Uncertainty Associated with Prompt Analysis
Accurate determination of the properties of biomass is of particular interest in studies on biomass combustion or cofiring. The aim of this paper is to develop a methodology for prompt analysis of heterogeneous solid fuels with an acceptable degree of accuracy. Special care must be taken with the sampling procedure to achieve an acceptable degree of error and low statistical uncertainty. A samp...
متن کاملCombined uncertainty factor for sampling and analysis
Measurement uncertainty that arises from primary sampling can be expressed as an uncertainty factor, which recognises its sometimes approximately log-normal probability distribution. By contrast, uncertainty arising from chemical analysis is usually expressed as relative uncertainty, based upon the assumptions of its approximately normal distribution. A new method is proposed that enables uncer...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- BMJ
دوره 349 شماره
صفحات -
تاریخ انتشار 2014